# Knowledge distillation optimization

Xlm Roberta Ua Distilled
MIT
This is a fine-tuned sentence transformer model based on xlm-roberta-base, supporting English and Ukrainian, suitable for tasks like semantic textual similarity and semantic search.
Text Embedding Supports Multiple Languages
X
panalexeu
121
1
Codev R1 Distill Qwen 7B
A Verilog RTL code generation model distilled from DeepSeek-R1, demonstrating outstanding performance in Verilog benchmarks
Large Language Model Transformers
C
zhuyaoyu
154
2
Aimv2 Large Patch14 336.apple Pt Dist
AIM-v2 is an efficient image encoder implemented based on the timm library, suitable for various computer vision tasks.
Image Classification Transformers
A
timm
14
0
Ultravox V0 4
MIT
Ultravox is a multimodal voice large language model based on Llama3.1-8B-Instruct and Whisper-medium, capable of processing both voice and text inputs simultaneously.
Audio-to-Text Transformers Supports Multiple Languages
U
fixie-ai
1,851
48
Distilbert PoliticalBias
MIT
A fine-tuned model based on DistilBERT for detecting and reducing political bias in text, utilizing knowledge distillation and diffusion techniques to achieve unbiased text representation.
Text Classification Transformers English
D
cajcodes
265
2
Multilingual Distilwhisper 28k
MIT
An improved multilingual automatic speech recognition model based on whisper-small, enhancing target language performance through CLSR module and knowledge distillation
Speech Recognition Transformers Other
M
naver
47
13
Lamini T5 738M
LaMini-T5-738M is an instruction-tuned model based on t5-large, fine-tuned on the LaMini-instruction dataset, with 738M parameters, and is one of the LaMini-LM series models.
Large Language Model Transformers English
L
MBZUAI
2,966
49
Semantic Xlmr Bn
A multilingual sentence embedding model optimized for Bengali, mapping text to a 768-dimensional vector space
Text Embedding Transformers Other
S
afschowdhury
225
1
Dynamic Minilmv2 L6 H384 Squad1.1 Int8 Static
MIT
QuaLA-MiniLM is a compact language model developed by Intel, integrating knowledge distillation, length-adaptive transformers, and 8-bit quantization technology. It achieves up to 8.8x acceleration on the SQuAD1.1 dataset with less than 1% accuracy loss.
Large Language Model Transformers
D
Intel
172
0
Minilmv2 L6 H384 Distilled From RoBERTa Large
MiniLMv2 is a lightweight language representation model developed by Microsoft, achieving efficient performance through knowledge distillation techniques.
Large Language Model Transformers
M
nreimers
73
6
Minilmv2 L6 H384 Distilled From BERT Large
MiniLMv2 is a lightweight language representation model developed by Microsoft, achieving efficient inference through knowledge distillation techniques, suitable for various natural language processing tasks.
Large Language Model Transformers
M
nreimers
14.21k
1
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase